Gaussian-Dirichlet Posterior Dominance in Sequential Learning
نویسندگان
چکیده
We consider the problem of sequential learning from categorical observations bounded in [0, 1]. We establish an ordering between the Dirichlet posterior over categorical outcomes and a Gaussian posterior under observations with N(0, 1) noise. We establish that, conditioned upon identical data with at least two observations, the posterior mean of the categorical distribution will always second-order stochastically dominate the posterior mean of the Gaussian distribution. These results provide a useful tool for the analysis of sequential learning under categorical outcomes.
منابع مشابه
Transfer Learning for Reinforcement Learning with Dependent Dirichlet Process and Gaussian Process
The ability to transfer knowledge across tasks is important in guaranteeing the performance of lifelong learning in autonomous agents. We propose a flexible Bayesian Nonparametric (BNP) model based architecture for transferring knowledge between reinforcement learning domains. A Dependent Dirichlet Process Gaussian Process hierarchial BNP model is used to cluster different classes of source MDP...
متن کاملApproximate Inference for Deep Latent Gaussian Mixtures
Deep latent Gaussian models (DLGMs) composed of density and inference networks [14]—the pipeline that defines a Variational Autoencoder [8]—have achieved notable success on tasks ranging from image modeling [3] to semi-supervised classification [6, 11]. However, the approximate posterior in these models is usually chosen to be a factorized Gaussian, thereby imposing strong constraints on the po...
متن کاملBayes Meets Bellman: The Gaussian Process Approach to Temporal Difference Learning
We present a novel Bayesian approach to the problem of value function estimation in continuous state spaces. We define a probabilistic generative model for the value function by imposing a Gaussian prior over value functions and assuming a Gaussian noise model. Due to the Gaussian nature of the random processes involved, the posterior distribution of the value function is also Gaussian and is t...
متن کاملCharacterizing the Function Space for Bayesian Kernel Models
Kernel methods have been very popular in the machine learning literature in the last ten years, mainly in the context of Tikhonov regularization algorithms. In this paper we study a coherent Bayesian kernel model based on an integral operator defined as the convolution of a kernel with a signed measure. Priors on the random signed measures correspond to prior distributions on the functions mapp...
متن کاملAdaptive Low-Complexity Sequential Inference for Dirichlet Process Mixture Models
We develop a sequential low-complexity inference procedure for the Infinite Gaussian Mixture Model (IGMM) for the general case of an unknown mean and covariance. The observations are sequentially allocated to classes based on a sequential maximum a-posterior (MAP) criterion. We present an easily computed, closed form for the conditional likelihood, in which the parameters can be recursively upd...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- CoRR
دوره abs/1702.04126 شماره
صفحات -
تاریخ انتشار 2017